Remove unused 'retry' parameter from on_selected_cpus() etc.
authorKeir Fraser <keir.fraser@citrix.com>
Wed, 27 May 2009 10:16:27 +0000 (11:16 +0100)
committerKeir Fraser <keir.fraser@citrix.com>
Wed, 27 May 2009 10:16:27 +0000 (11:16 +0100)
Remove the unused "retry" parameter of on_selected_cpus(),
on_each_cpu(), smp_call_function(), and smp_call_function_single().

Signed-off-by: Jan Beulich <jbeulich@novell.com>
34 files changed:
xen/arch/ia64/linux-xen/mca.c
xen/arch/ia64/linux-xen/perfmon.c
xen/arch/ia64/linux-xen/smp.c
xen/arch/ia64/linux-xen/smpboot.c
xen/arch/ia64/linux-xen/sn/kernel/sn2_smp.c
xen/arch/ia64/vmx/vmmu.c
xen/arch/ia64/vmx/vtlb.c
xen/arch/ia64/xen/cpufreq/cpufreq.c
xen/arch/ia64/xen/flushtlb.c
xen/arch/ia64/xen/fw_emul.c
xen/arch/ia64/xen/vhpt.c
xen/arch/x86/acpi/cpufreq/cpufreq.c
xen/arch/x86/acpi/cpufreq/powernow.c
xen/arch/x86/cpu/amd.c
xen/arch/x86/cpu/mcheck/amd_nonfatal.c
xen/arch/x86/cpu/mcheck/mce.c
xen/arch/x86/cpu/mcheck/mce_intel.c
xen/arch/x86/cpu/mcheck/non-fatal.c
xen/arch/x86/cpu/mtrr/main.c
xen/arch/x86/hvm/hvm.c
xen/arch/x86/hvm/svm/svm.c
xen/arch/x86/hvm/vmx/vmcs.c
xen/arch/x86/hvm/vmx/vmx.c
xen/arch/x86/irq.c
xen/arch/x86/machine_kexec.c
xen/arch/x86/oprofile/nmi_int.c
xen/arch/x86/shutdown.c
xen/arch/x86/smp.c
xen/arch/x86/time.c
xen/arch/x86/x86_32/traps.c
xen/common/gdbstub.c
xen/common/keyhandler.c
xen/include/asm-ia64/linux-xen/asm/smp.h
xen/include/xen/smp.h

index 8e9d9f74015ebf962463696fec1c46d009aefbd9..a0ba2fffe918670909a3a70ca0010635a4a510f7 100644 (file)
@@ -956,7 +956,7 @@ ia64_mca_cmc_vector_enable (void *dummy)
 static void
 ia64_mca_cmc_vector_disable_keventd(void *unused)
 {
-       on_each_cpu(ia64_mca_cmc_vector_disable, NULL, 1, 0);
+       on_each_cpu(ia64_mca_cmc_vector_disable, NULL, 0);
 }
 
 /*
@@ -968,7 +968,7 @@ ia64_mca_cmc_vector_disable_keventd(void *unused)
 static void
 ia64_mca_cmc_vector_enable_keventd(void *unused)
 {
-       on_each_cpu(ia64_mca_cmc_vector_enable, NULL, 1, 0);
+       on_each_cpu(ia64_mca_cmc_vector_enable, NULL, 0);
 }
 #endif /* !XEN */
 
index c905afc2b55db48ede886abb38119127d0d9f982..6aed1614087c44568881cbdf75a3b46e9b506c0d 100644 (file)
@@ -1895,7 +1895,7 @@ pfm_syswide_cleanup_other_cpu(pfm_context_t *ctx)
        int ret;
 
        DPRINT(("calling CPU%d for cleanup\n", ctx->ctx_cpu));
-       ret = smp_call_function_single(ctx->ctx_cpu, pfm_syswide_force_stop, ctx, 0, 1);
+       ret = smp_call_function_single(ctx->ctx_cpu, pfm_syswide_force_stop, ctx, 1);
        DPRINT(("called CPU%d for cleanup ret=%d\n", ctx->ctx_cpu, ret));
 }
 #endif /* CONFIG_SMP */
@@ -6895,7 +6895,7 @@ pfm_install_alt_pmu_interrupt(pfm_intr_handler_desc_t *hdl)
        }
 
        /* save the current system wide pmu states */
-       ret = on_each_cpu(pfm_alt_save_pmu_state, NULL, 0, 1);
+       ret = on_each_cpu(pfm_alt_save_pmu_state, NULL, 1);
        if (ret) {
                DPRINT(("on_each_cpu() failed: %d\n", ret));
                goto cleanup_reserve;
@@ -6940,7 +6940,7 @@ pfm_remove_alt_pmu_interrupt(pfm_intr_handler_desc_t *hdl)
 
        pfm_alt_intr_handler = NULL;
 
-       ret = on_each_cpu(pfm_alt_restore_pmu_state, NULL, 0, 1);
+       ret = on_each_cpu(pfm_alt_restore_pmu_state, NULL, 1);
        if (ret) {
                DPRINT(("on_each_cpu() failed: %d\n", ret));
        }
@@ -7499,7 +7499,7 @@ xenpfm_context_load(XEN_GUEST_HANDLE(pfarg_load_t) req)
 
        BUG_ON(in_irq());
        spin_lock(&xenpfm_context_lock);
-       smp_call_function(&xenpfm_context_load_cpu, &arg, 1, 1);
+       smp_call_function(&xenpfm_context_load_cpu, &arg, 1);
        xenpfm_context_load_cpu(&arg);
        spin_unlock(&xenpfm_context_lock);
        for_each_online_cpu(cpu) {
@@ -7553,7 +7553,7 @@ xenpfm_context_unload(void)
                return error;
        }
 
-       smp_call_function(&xenpfm_context_unload_cpu, &arg, 1, 1);
+       smp_call_function(&xenpfm_context_unload_cpu, &arg, 1);
        xenpfm_context_unload_cpu(&arg);
        spin_unlock(&xenpfm_context_lock);
        for_each_online_cpu(cpu) {
index 03d673f1c2eca7260019b6fa8080d479597ad3b0..d5bf978a61665d6499b58ddb0ad911d084dca364 100644 (file)
@@ -274,7 +274,7 @@ smp_send_reschedule (int cpu)
 void
 smp_flush_tlb_all (void)
 {
-       on_each_cpu((void (*)(void *))local_flush_tlb_all, NULL, 1, 1);
+       on_each_cpu((void (*)(void *))local_flush_tlb_all, NULL, 1);
 }
 
 void
@@ -297,7 +297,7 @@ smp_flush_tlb_mm (struct mm_struct *mm)
         * anyhow, and once a CPU is interrupted, the cost of local_flush_tlb_all() is
         * rather trivial.
         */
-       on_each_cpu((void (*)(void *))local_finish_flush_tlb_mm, mm, 1, 1);
+       on_each_cpu((void (*)(void *))local_finish_flush_tlb_mm, mm, 1);
 }
 #endif
 
@@ -314,7 +314,7 @@ smp_flush_tlb_mm (struct mm_struct *mm)
  */
 
 int
-smp_call_function_single (int cpuid, void (*func) (void *info), void *info, int nonatomic,
+smp_call_function_single (int cpuid, void (*func) (void *info), void *info,
                          int wait)
 {
        struct call_data_struct data;
@@ -372,7 +372,6 @@ EXPORT_SYMBOL(smp_call_function_single);
  *  [SUMMARY]  Run a function on all other CPUs.
  *  <func>     The function to run. This must be fast and non-blocking.
  *  <info>     An arbitrary pointer to pass to the function.
- *  <nonatomic>        currently unused.
  *  <wait>     If true, wait (atomically) until function has completed on other CPUs.
  *  [RETURNS]   0 on success, else a negative status code.
  *
@@ -383,7 +382,7 @@ EXPORT_SYMBOL(smp_call_function_single);
  * hardware interrupt handler or from a bottom half handler.
  */
 int
-smp_call_function (void (*func) (void *info), void *info, int nonatomic, int wait)
+smp_call_function (void (*func) (void *info), void *info, int wait)
 {
        struct call_data_struct data;
        int cpus = num_online_cpus()-1;
@@ -438,7 +437,7 @@ EXPORT_SYMBOL(smp_call_function);
 #ifdef XEN
 int
 on_selected_cpus(const cpumask_t *selected, void (*func) (void *info),
-                 void *info, int retry, int wait)
+                 void *info, int wait)
 {
        struct call_data_struct data;
        unsigned int cpu, nr_cpus = cpus_weight(*selected);
index a450dec730b47caf5a88539d85d4044d6b5c1863..38f761796dff76dda46e3b7a2e51df466245aa7d 100644 (file)
@@ -307,7 +307,7 @@ ia64_sync_itc (unsigned int master)
 
        go[MASTER] = 1;
 
-       if (smp_call_function_single(master, sync_master, NULL, 1, 0) < 0) {
+       if (smp_call_function_single(master, sync_master, NULL, 0) < 0) {
                printk(KERN_ERR "sync_itc: failed to get attention of CPU %u!\n", master);
                return;
        }
index b525c8b41183a14dc0198e12b518efbb1afd9d16..4626b57187f9366d6dcfc86ab9cf7de50d96fad5 100644 (file)
@@ -240,7 +240,7 @@ sn2_global_tlb_purge(unsigned long start,
                flush_data.end = end;
                flush_data.nbits = nbits;
                on_selected_cpus(&selected_cpus, sn_flush_ptcga_cpu,
-                                &flush_data, 1, 1);
+                                &flush_data, 1);
        }
        spin_unlock(&sn2_ptcg_lock2);
 }
index 29f73be647cb5f1466e74b884d1547356411be7d..d61bde03a62803e541e0776b1a0980a9a869aad8 100644 (file)
@@ -448,8 +448,7 @@ IA64FAULT vmx_vcpu_ptc_ga(VCPU *vcpu, u64 va, u64 ps)
             if (cpu != current->processor) {
                 spin_barrier(&per_cpu(schedule_data, cpu).schedule_lock);
                 /* Flush VHPT on remote processors. */
-                smp_call_function_single(cpu, &ptc_ga_remote_func,
-                                         &args, 0, 1);
+                smp_call_function_single(cpu, &ptc_ga_remote_func, &args, 1);
             } else {
                 ptc_ga_remote_func(&args);
             }
index 74b0da7f41e09ebc380f7df501de398beaa12bf7..9f16e45914e2756380cadc4b1daff6cd77bef63c 100644 (file)
@@ -643,7 +643,7 @@ void vmx_vcpu_flush_vtlb_all(VCPU *v)
     if (v->processor == smp_processor_id())
         __thash_purge_all(v);
     else
-        smp_call_function_single(v->processor, __thash_purge_all, v, 1, 1);
+        smp_call_function_single(v->processor, __thash_purge_all, v, 1);
     vcpu_unpause(v);
 }
 
index fdd688810823a75a189893e7590c8fa1c7251a0c..b365e61d943bb8ea27d1cb5bbbe3d8a06edcc52b 100644 (file)
@@ -95,8 +95,7 @@ acpi_cpufreq_get (unsigned int cpu)
        if (cpu == smp_processor_id())
                processor_get_freq((void*)&freq);
        else
-               smp_call_function_single(cpu, processor_get_freq,
-                                        (void *)&freq, 0, 1);
+               smp_call_function_single(cpu, processor_get_freq, &freq, 1);
 
        return freq;
 }
@@ -143,8 +142,7 @@ processor_set_freq (struct acpi_cpufreq_data *data,
        if (cpu == smp_processor_id())
                processor_set_pstate((void *)&value);
        else
-               smp_call_function_single(cpu, processor_set_pstate,
-                               (void *)&value, 0, 1);
+               smp_call_function_single(cpu, processor_set_pstate, &value, 1);
 
        if (value) {
                printk(KERN_WARNING "Transition failed\n");
index 19aabf6c50547470ae550e022bdca0584cbf59c1..e8ff5943848166d9b33a7c6ab5247c756a25dd08 100644 (file)
@@ -70,7 +70,7 @@ void
 new_tlbflush_clock_period(void)
 {
     /* flush all vhpt of physical cpu and mTLB */
-    on_each_cpu(tlbflush_clock_local_flush, NULL, 1, 1);
+    on_each_cpu(tlbflush_clock_local_flush, NULL, 1);
 
     /*
      * if global TLB shootdown is finished, increment tlbflush_time
index 2c9235a38cb4ad3377e271e68620b94c1a4a56fe..c4b0c0327b3c683c77c9c1769f41dfae1146e0b7 100644 (file)
@@ -281,7 +281,7 @@ sal_emulator (long index, unsigned long in1, unsigned long in2,
                                IA64_SAL_DEBUG("SAL_GET_STATE_INFO: remote\n");
                                ret = smp_call_function_single(e->cpuid,
                                                               get_state_info_on,
-                                                              &arg, 0, 1);
+                                                              &arg, 1);
                                if (ret < 0) {
                                        printk("SAL_GET_STATE_INFO "
                                               "smp_call_function_single error:"
@@ -344,7 +344,7 @@ sal_emulator (long index, unsigned long in1, unsigned long in2,
                                int ret;
                                IA64_SAL_DEBUG("SAL_CLEAR_STATE_INFO: remote\n");
                                ret = smp_call_function_single(e->cpuid,
-                                       clear_state_info_on, &arg, 0, 1);
+                                       clear_state_info_on, &arg, 1);
                                if (ret < 0) {
                                        printk("sal_emulator: "
                                               "SAL_CLEAR_STATE_INFO "
@@ -845,8 +845,7 @@ xen_pal_emulator(unsigned long index, u64 in1, u64 in2, u64 in3)
                                .progress = 0,
                                .status = 0
                        };
-                       smp_call_function(remote_pal_cache_flush,
-                                         (void *)&args, 1, 1);
+                       smp_call_function(remote_pal_cache_flush, &args, 1);
                        if (args.status != 0)
                                panic_domain(NULL, "PAL_CACHE_FLUSH ERROR, "
                                             "remote status %lx", args.status);
@@ -945,7 +944,7 @@ xen_pal_emulator(unsigned long index, u64 in1, u64 in2, u64 in3)
                        /* must be performed on all remote processors 
                           in the coherence domain. */
                        smp_call_function(remote_pal_prefetch_visibility,
-                                         (void *)in1, 1, 1);
+                                         (void *)in1, 1);
                        status = 1; /* no more necessary on remote processor */
                }
                break;
@@ -953,7 +952,7 @@ xen_pal_emulator(unsigned long index, u64 in1, u64 in2, u64 in3)
                status = ia64_pal_mc_drain();
                /* FIXME: All vcpus likely call PAL_MC_DRAIN.
                   That causes the congestion. */
-               smp_call_function(remote_pal_mc_drain, NULL, 1, 1);
+               smp_call_function(remote_pal_mc_drain, NULL, 1);
                break;
            case PAL_BRAND_INFO:
                if (in1 == 0) {
index 212f5509606e09030d8408ce21689b39c4cc8823..a7648fc7d1fcf7227a5d174ca2b6b1cf4837b0ac 100644 (file)
@@ -307,7 +307,7 @@ void domain_flush_vtlb_all(struct domain* d)
                        // takes care of mTLB flush.
                        smp_call_function_single(v->processor,
                                                 __vcpu_flush_vtlb_all,
-                                                v, 1, 1);
+                                                v, 1);
        }
        perfc_incr(domain_flush_vtlb_all);
 }
@@ -513,9 +513,9 @@ void domain_flush_tlb_vhpt(struct domain *d)
 {
        /* Very heavy...  */
        if (HAS_PERVCPU_VHPT(d) || is_hvm_domain(d))
-               on_each_cpu((void (*)(void *))local_flush_tlb_all, NULL, 1, 1);
+               on_each_cpu((void (*)(void *))local_flush_tlb_all, NULL, 1);
        else
-               on_each_cpu((void (*)(void *))flush_tlb_vhpt_all, d, 1, 1);
+               on_each_cpu((void (*)(void *))flush_tlb_vhpt_all, d, 1);
        cpus_clear (d->domain_dirty_cpumask);
 }
 
@@ -532,7 +532,7 @@ void flush_tlb_for_log_dirty(struct domain *d)
                        thash_purge_all(v);
                }
                smp_call_function((void (*)(void *))local_flush_tlb_all, 
-                                       NULL, 1, 1);
+                                       NULL, 1);
        } else if (HAS_PERVCPU_VHPT(d)) {
                for_each_vcpu (d, v) {
                        if (!v->is_initialised)
@@ -541,9 +541,9 @@ void flush_tlb_for_log_dirty(struct domain *d)
                        vcpu_purge_tr_entry(&PSCBX(v,itlb));
                        vcpu_vhpt_flush(v);
                }
-               on_each_cpu((void (*)(void *))local_flush_tlb_all, NULL, 1, 1);
+               on_each_cpu((void (*)(void *))local_flush_tlb_all, NULL, 1);
        } else {
-               on_each_cpu((void (*)(void *))flush_tlb_vhpt_all, d, 1, 1);
+               on_each_cpu((void (*)(void *))flush_tlb_vhpt_all, d, 1);
        }
        cpus_clear (d->domain_dirty_cpumask);
 }
@@ -562,7 +562,7 @@ void flush_tlb_mask(const cpumask_t *mask)
     for_each_cpu_mask (cpu, *mask)
         if (cpu != smp_processor_id())
             smp_call_function_single
-                (cpu, (void (*)(void *))flush_tlb_vhpt_all, NULL, 1, 1);
+                (cpu, (void (*)(void *))flush_tlb_vhpt_all, NULL, 1);
 }
 
 #ifdef PERF_COUNTERS
index ccf7bffa756a46ea924408dd176ffd094fca7c4c..8408f1669e95a9f033791744718d61e18396845c 100644 (file)
@@ -186,7 +186,7 @@ static void drv_read(struct drv_cmd *cmd)
     if (likely(cpu_isset(smp_processor_id(), cmd->mask)))
         do_drv_read((void *)cmd);
     else
-        on_selected_cpus(&cmd->mask, do_drv_read, (void *)cmd, 0, 1);
+        on_selected_cpus(&cmd->mask, do_drv_read, cmd, 1);
 }
 
 static void drv_write(struct drv_cmd *cmd)
@@ -195,7 +195,7 @@ static void drv_write(struct drv_cmd *cmd)
         cpu_isset(smp_processor_id(), cmd->mask))
         do_drv_write((void *)cmd);
     else
-        on_selected_cpus(&cmd->mask, do_drv_write, (void *)cmd, 0, 0);
+        on_selected_cpus(&cmd->mask, do_drv_write, cmd, 0);
 }
 
 static u32 get_cur_val(cpumask_t mask)
@@ -303,7 +303,7 @@ static unsigned int get_measured_perf(unsigned int cpu, unsigned int flag)
         read_measured_perf_ctrs((void *)&readin);
     } else {
         on_selected_cpus(cpumask_of(cpu), read_measured_perf_ctrs,
-                        (void *)&readin, 0, 1);
+                        &readin, 1);
     }
 
     cur.aperf.whole = readin.aperf.whole - saved->aperf.whole;
index 98b8ed2df62611feb4f9ed005fea008b9373fa4e..610a28e3eab54694592ca102f5e0c66133a217fe 100644 (file)
@@ -121,7 +121,7 @@ static int powernow_cpufreq_target(struct cpufreq_policy *policy,
 
     cmd.val = next_perf_state;
 
-    on_selected_cpus(&cmd.mask, transition_pstate, (void *) &cmd, 0, 0);
+    on_selected_cpus(&cmd.mask, transition_pstate, &cmd, 0);
 
     perf->state = next_perf_state;
     policy->cur = freqs.new;
index f2ff228a16379335f0c8b3e0e30cc111cc9fa3ee..48b8cccd20ddf5f146bf79f101f09282509cade3 100644 (file)
@@ -246,7 +246,7 @@ static void check_disable_c1e(unsigned int port, u8 value)
 {
        /* C1E is sometimes enabled during entry to ACPI mode. */
        if ((port == acpi_smi_cmd) && (value == acpi_enable_value))
-               on_each_cpu(disable_c1e, NULL, 1, 1);
+               on_each_cpu(disable_c1e, NULL, 1);
 }
 
 static void __devinit init_amd(struct cpuinfo_x86 *c)
index d354b1f06b2bfefa3ef074c63593be1fd9410117..1861d16e43c690a9714e6dac08d16e2e6edd9c2d 100644 (file)
@@ -133,7 +133,7 @@ void mce_amd_checkregs(void *info)
  */
 static void mce_amd_work_fn(void *data)
 {
-       on_each_cpu(mce_amd_checkregs, data, 1, 1);
+       on_each_cpu(mce_amd_checkregs, data, 1);
 
        if (adjust > 0) {
                if (!guest_enabled_event(dom0->vcpu[0], VIRQ_MCA) ) {
index 655c168d4256d93f68b3fb9beb6cd5b06235d786..2ef204315e05fa19f81ef6ee27e1f9d6c47696d6 100644 (file)
@@ -1162,8 +1162,7 @@ long do_mca(XEN_GUEST_HANDLE(xen_mc_t) u_xen_mc)
                        if (log_cpus == NULL)
                                return x86_mcerr("do_mca cpuinfo", -ENOMEM);
 
-                       if (on_each_cpu(do_mc_get_cpu_info, log_cpus,
-                           1, 1) != 0) {
+                       if (on_each_cpu(do_mc_get_cpu_info, log_cpus, 1)) {
                                xfree(log_cpus);
                                return x86_mcerr("do_mca cpuinfo", -EIO);
                        }
@@ -1206,7 +1205,7 @@ long do_mca(XEN_GUEST_HANDLE(xen_mc_t) u_xen_mc)
                add_taint(TAINT_ERROR_INJECT);
 
                on_selected_cpus(cpumask_of(target), x86_mc_msrinject,
-                                mc_msrinject, 1, 1);
+                                mc_msrinject, 1);
 
                break;
 
@@ -1226,7 +1225,7 @@ long do_mca(XEN_GUEST_HANDLE(xen_mc_t) u_xen_mc)
                add_taint(TAINT_ERROR_INJECT);
 
                on_selected_cpus(cpumask_of(target), x86_mc_mceinject,
-                                mc_mceinject, 1, 1);
+                                mc_mceinject, 1);
                break;
 
        default:
index 6bf8c87addf260ed57ecdec77e29461e993cc0d9..f3aeac8838a26b0dd87989fd61fc09d8d8d30d32 100644 (file)
@@ -632,7 +632,7 @@ static void __cpu_mcheck_distribute_cmci(void *unused)
 void cpu_mcheck_distribute_cmci(void)
 {
     if (cmci_support && !mce_disabled)
-        on_each_cpu(__cpu_mcheck_distribute_cmci, NULL, 0, 0);
+        on_each_cpu(__cpu_mcheck_distribute_cmci, NULL, 0);
 }
 
 static void clear_cmci(void)
index 9d66a12b6b981947c749338267eed05552353913..bae8d06803c83d0e2375a5e8b008821a5bcd6078 100644 (file)
@@ -69,7 +69,7 @@ static void mce_checkregs (void *info)
 
 static void mce_work_fn(void *data)
 { 
-       on_each_cpu(mce_checkregs, NULL, 1, 1);
+       on_each_cpu(mce_checkregs, NULL, 1);
 
        if (variable_period) {
                if (adjust)
index 0f15eb8688f34b2f0440401feae2e2cf4f735457..44e1c6d4aeb063d28bc460f7f9886a66ce190ca9 100644 (file)
@@ -229,7 +229,7 @@ static void set_mtrr(unsigned int reg, unsigned long base,
        atomic_set(&data.gate,0);
 
        /*  Start the ball rolling on other CPUs  */
-       if (smp_call_function(ipi_handler, &data, 1, 0) != 0)
+       if (smp_call_function(ipi_handler, &data, 0) != 0)
                panic("mtrr: timed out waiting for other CPUs\n");
 
        local_irq_save(flags);
@@ -688,7 +688,7 @@ void mtrr_save_state(void)
        if (cpu == 0)
                mtrr_save_fixed_ranges(NULL);
        else
-               on_selected_cpus(cpumask_of(0), mtrr_save_fixed_ranges, NULL, 1, 1);
+               on_selected_cpus(cpumask_of(0), mtrr_save_fixed_ranges, NULL, 1);
        put_cpu();
 }
 
index 6c9fadca0d08c180f543da0193c57a828d4269e3..fc31b3b1bda12fd8dc96787d27329b88f0aec055 100644 (file)
@@ -971,7 +971,7 @@ int hvm_set_cr0(unsigned long value)
             if ( !v->domain->arch.hvm_domain.is_in_uc_mode )
             {
                 /* Flush physical caches. */
-                on_each_cpu(local_flush_cache, NULL, 1, 1);
+                on_each_cpu(local_flush_cache, NULL, 1);
                 hvm_set_uc_mode(v, 1);
             }
             spin_unlock(&v->domain->arch.hvm_domain.uc_lock);
index d800e8cd1b2c51418b9be4b03880f9aec211d953..505a5149c62067fe7b684c41b257888430d7e9d2 100644 (file)
@@ -1254,7 +1254,7 @@ static void wbinvd_ipi(void *info)
 static void svm_wbinvd_intercept(void)
 {
     if ( has_arch_pdevs(current->domain) )
-        on_each_cpu(wbinvd_ipi, NULL, 1, 1);
+        on_each_cpu(wbinvd_ipi, NULL, 1);
 }
 
 static void svm_vmexit_do_invalidate_cache(struct cpu_user_regs *regs)
index 786a22796e037343f05a3747235cce86a943ddd6..400d31de74ea64ab266a19b98f3b514fa208c8a5 100644 (file)
@@ -264,7 +264,7 @@ static void vmx_clear_vmcs(struct vcpu *v)
     int cpu = v->arch.hvm_vmx.active_cpu;
 
     if ( cpu != -1 )
-        on_selected_cpus(cpumask_of(cpu), __vmx_clear_vmcs, v, 1, 1);
+        on_selected_cpus(cpumask_of(cpu), __vmx_clear_vmcs, v, 1);
 }
 
 static void vmx_load_vmcs(struct vcpu *v)
@@ -900,7 +900,7 @@ void vmx_do_resume(struct vcpu *v)
         {
             int cpu = v->arch.hvm_vmx.active_cpu;
             if ( cpu != -1 )
-                on_selected_cpus(cpumask_of(cpu), wbinvd_ipi, NULL, 1, 1);
+                on_selected_cpus(cpumask_of(cpu), wbinvd_ipi, NULL, 1);
         }
 
         vmx_clear_vmcs(v);
index c4eee9b44e08764dd6947f53b05ca0c0677dada3..dddd9ec108e2a60d757eacdefa7312dc95dbc662 100644 (file)
@@ -1220,7 +1220,7 @@ void ept_sync_domain(struct domain *d)
     if ( d->arch.hvm_domain.hap_enabled && d->vcpu[0] )
     {
         ASSERT(local_irq_is_enabled());
-        on_each_cpu(__ept_sync_domain, d, 1, 1);
+        on_each_cpu(__ept_sync_domain, d, 1);
     }
 }
 
@@ -2131,7 +2131,7 @@ static void vmx_wbinvd_intercept(void)
         return;
 
     if ( cpu_has_wbinvd_exiting )
-        on_each_cpu(wbinvd_ipi, NULL, 1, 1);
+        on_each_cpu(wbinvd_ipi, NULL, 1);
     else
         wbinvd();
 }
index 65e69001cddd51ba20f6d0e03b66413135667c86..8abaab913ec6da5c7d4537262df17c8bad6efceb 100644 (file)
@@ -522,7 +522,7 @@ static void __pirq_guest_eoi(struct domain *d, int irq)
     }
 
     if ( !cpus_empty(cpu_eoi_map) )
-        on_selected_cpus(&cpu_eoi_map, set_eoi_ready, desc, 1, 0);
+        on_selected_cpus(&cpu_eoi_map, set_eoi_ready, desc, 0);
 }
 
 int pirq_guest_eoi(struct domain *d, int irq)
@@ -761,7 +761,7 @@ static irq_guest_action_t *__pirq_guest_unbind(
         {
             cpu_eoi_map = action->cpu_eoi_map;
             spin_unlock_irq(&desc->lock);
-            on_selected_cpus(&cpu_eoi_map, set_eoi_ready, desc, 1, 0);
+            on_selected_cpus(&cpu_eoi_map, set_eoi_ready, desc, 0);
             spin_lock_irq(&desc->lock);
         }
         break;
@@ -799,7 +799,7 @@ static irq_guest_action_t *__pirq_guest_unbind(
     {
         BUG_ON(action->ack_type != ACKTYPE_EOI);
         spin_unlock_irq(&desc->lock);
-        on_selected_cpus(&cpu_eoi_map, set_eoi_ready, desc, 1, 1);
+        on_selected_cpus(&cpu_eoi_map, set_eoi_ready, desc, 1);
         spin_lock_irq(&desc->lock);
     }
 
index af7663e3fcb5ca6c885a71c4504581a8051eb629..748883459342f242104f1fb3f5ef8748ee1df0d6 100644 (file)
@@ -100,7 +100,7 @@ void machine_reboot_kexec(xen_kexec_image_t *image)
     if ( reboot_cpu_id != smp_processor_id() )
     {
         on_selected_cpus(cpumask_of(reboot_cpu_id), __machine_reboot_kexec,
-                         image, 1, 0);
+                         image, 0);
         for (;;)
                 ; /* nothing */
     }
index c01adab4ab116d9ef841e0fb53ea460d8b21a491..89040890d688d68b5c3e55e9541a51e5fff8b01e 100644 (file)
@@ -186,7 +186,7 @@ static void nmi_cpu_setup(void * dummy)
 
 int nmi_setup_events(void)
 {
-       on_each_cpu(nmi_cpu_setup, NULL, 0, 1);
+       on_each_cpu(nmi_cpu_setup, NULL, 1);
        return 0;
 }
 
@@ -207,7 +207,7 @@ int nmi_reserve_counters(void)
        /* We need to serialize save and setup for HT because the subset
         * of msrs are distinct for save and setup operations
         */
-       on_each_cpu(nmi_save_registers, NULL, 0, 1);
+       on_each_cpu(nmi_save_registers, NULL, 1);
        return 0;
 }
 
@@ -256,7 +256,7 @@ static void nmi_cpu_shutdown(void * dummy)
  
 void nmi_release_counters(void)
 {
-       on_each_cpu(nmi_cpu_shutdown, NULL, 0, 1);
+       on_each_cpu(nmi_cpu_shutdown, NULL, 1);
        release_lapic_nmi();
        free_msrs();
 }
@@ -274,7 +274,7 @@ static void nmi_cpu_start(void * dummy)
 
 int nmi_start(void)
 {
-       on_each_cpu(nmi_cpu_start, NULL, 0, 1);
+       on_each_cpu(nmi_cpu_start, NULL, 1);
        return 0;
 }
  
@@ -306,7 +306,7 @@ static void nmi_cpu_stop(void * dummy)
  
 void nmi_stop(void)
 {
-       on_each_cpu(nmi_cpu_stop, NULL, 0, 1);
+       on_each_cpu(nmi_cpu_stop, NULL, 1);
 }
 
 
index 1d8218fa27b3a22f9a888e257c111af1e9cf9465..15f5988af7874273ed7daa54c67ff42110ca628a 100644 (file)
@@ -91,7 +91,7 @@ void machine_halt(void)
     watchdog_disable();
     console_start_sync();
     local_irq_enable();
-    smp_call_function(__machine_halt, NULL, 1, 0);
+    smp_call_function(__machine_halt, NULL, 0);
     __machine_halt(NULL);
 }
 
@@ -311,7 +311,7 @@ void machine_restart(unsigned int delay_millisecs)
     {
         /* Send IPI to the boot CPU (logical cpu 0). */
         on_selected_cpus(cpumask_of(0), __machine_restart,
-                         &delay_millisecs, 1, 0);
+                         &delay_millisecs, 0);
         for ( ; ; )
             halt();
     }
index 0bf82da494daefe5ddf59cdb585e417597cb6ba7..9c1475051aeab99a01d4544fdeb672776ea8f16f 100644 (file)
@@ -239,19 +239,17 @@ static void __smp_call_function_interrupt(void);
 int smp_call_function(
     void (*func) (void *info),
     void *info,
-    int retry,
     int wait)
 {
     cpumask_t allbutself = cpu_online_map;
     cpu_clear(smp_processor_id(), allbutself);
-    return on_selected_cpus(&allbutself, func, info, retry, wait);
+    return on_selected_cpus(&allbutself, func, info, wait);
 }
 
 int on_selected_cpus(
     const cpumask_t *selected,
     void (*func) (void *info),
     void *info,
-    int retry,
     int wait)
 {
     struct call_data_struct data;
@@ -322,7 +320,7 @@ void smp_send_stop(void)
 {
     int timeout = 10;
 
-    smp_call_function(stop_this_cpu, NULL, 1, 0);
+    smp_call_function(stop_this_cpu, NULL, 0);
 
     /* Wait 10ms for all other CPUs to go offline. */
     while ( (num_online_cpus() > 1) && (timeout-- > 0) )
index 495ee844dc5be3c00f441d4848d96f40a73c3ed9..48d4ee1e9d7d1495a6608ae5fd10db3ddf0d42db 100644 (file)
@@ -1193,7 +1193,7 @@ static void time_calibration(void *unused)
                      opt_consistent_tscs
                      ? time_calibration_tsc_rendezvous
                      : time_calibration_std_rendezvous,
-                     &r, 0, 1);
+                     &r, 1);
 }
 
 void init_percpu_time(void)
index e3d716ec74730b19cfa852fbc33726ac3ee0e105..062c996d11e1bd7060285addee83722436550429 100644 (file)
@@ -403,7 +403,7 @@ static long register_guest_callback(struct callback_register *reg)
     case CALLBACKTYPE_sysenter_deprecated:
         if ( !cpu_has_sep )
             ret = -EINVAL;
-        else if ( on_each_cpu(do_update_sysenter, &reg->address, 1, 1) != 0 )
+        else if ( on_each_cpu(do_update_sysenter, &reg->address, 1) != 0 )
             ret = -EIO;
         break;
 
index e578756a7414ac87d698dec36bc33d2e921bbee1..6bbd23f395a08a48e5598e1fea587e45d951b64d 100644 (file)
@@ -672,7 +672,7 @@ static void gdb_smp_pause(void)
 
     atomic_set(&gdb_smp_paused_count, 0);
 
-    smp_call_function(gdb_pause_this_cpu, NULL, /* dont wait! */0, 0);
+    smp_call_function(gdb_pause_this_cpu, NULL, /* dont wait! */0);
 
     /* Wait 100ms for all other CPUs to enter pause loop */
     while ( (atomic_read(&gdb_smp_paused_count) < (num_online_cpus() - 1)) 
index 5407456290c615751640a29e67ec6b6d0684b6cb..a1fecf23b0f86d92f80b8154a6741c879359cef9 100644 (file)
@@ -119,7 +119,7 @@ static void dump_registers(unsigned char key, struct cpu_user_regs *regs)
         if ( cpu == smp_processor_id() )
             continue;
         printk("\n*** Dumping CPU%d host state: ***\n", cpu);
-        on_selected_cpus(cpumask_of(cpu), __dump_execstate, NULL, 1, 1);
+        on_selected_cpus(cpumask_of(cpu), __dump_execstate, NULL, 1);
     }
 
     printk("\n");
@@ -263,7 +263,7 @@ static void read_clocks(unsigned char key)
 
     spin_lock(&lock);
 
-    smp_call_function(read_clocks_slave, NULL, 0, 0);
+    smp_call_function(read_clocks_slave, NULL, 0);
 
     local_irq_disable();
     read_clocks_cpumask = cpu_online_map;
index e571562a88864dadcaadc5058f650fa32b3669fc..2831c0460c597e1a16c2b3c2d96e28ae878c12fc 100644 (file)
@@ -127,8 +127,8 @@ extern void __init smp_build_cpu_map(void);
 extern void __init init_smp_config (void);
 extern void smp_do_timer (struct pt_regs *regs);
 
-extern int smp_call_function_single (int cpuid, void (*func) (void *info), void *info,
-                                    int retry, int wait);
+extern int smp_call_function_single (int cpuid, void (*func) (void *info),
+                                    void *info, int wait);
 extern void smp_send_reschedule (int cpu);
 #ifdef XEN
 extern void lock_ipi_calllock(unsigned long *flags);
index 93b0713205d81eedb82e421ee4758fa9ad25b849..3d889ba66240bd1cd8fec11f1244375b4fcfbe34 100644 (file)
@@ -34,7 +34,6 @@ extern void smp_cpus_done(unsigned int max_cpus);
 extern int smp_call_function(
     void (*func) (void *info),
     void *info,
-    int retry,
     int wait);
 
 /* 
@@ -44,7 +43,6 @@ extern int on_selected_cpus(
     const cpumask_t *selected,
     void (*func) (void *info),
     void *info,
-    int retry,
     int wait);
 
 /*
@@ -59,10 +57,9 @@ void smp_prepare_boot_cpu(void);
 static inline int on_each_cpu(
     void (*func) (void *info),
     void *info,
-    int retry,
     int wait)
 {
-    return on_selected_cpus(&cpu_online_map, func, info, retry, wait);
+    return on_selected_cpus(&cpu_online_map, func, info, wait);
 }
 
 #define smp_processor_id() raw_smp_processor_id()